Change in Notation. e (i) (n) Note that this is inconsistent with the notation used earlier, c k+1x(n k) e (i) (n) x(n i) ˆx(n i) M

Size: px
Start display at page:

Download "Change in Notation. e (i) (n) Note that this is inconsistent with the notation used earlier, c k+1x(n k) e (i) (n) x(n i) ˆx(n i) M"

Transcription

1 Overview of Linear Prediction Terms and definitions Nonstationary case Stationary case Forward linear prediction Backward linear prediction Stationary processes Exchange matrices Examples Properties Introduction Important for more applications than just prediction Prominent role in spectral estimation, Kalman filtering, fast algorithms, etc Prediction is equivalent to whitening! (more later) Clearly many practical applications as well J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 1 J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 2 Nonstationary Problem Definition Given a segment of a signal {x(n),x(n 1),,x(n M)} of a stochastic process estimate x(n i) for i M using the remaining portion of the signal where c i (n) 1 ˆx(n i) c k(n)x(n k) k i e (i) (n) x(n i) ˆx(n i) = c k (n)x(n k) ˆx(n i) Change in Notation c k(n)x(n k) k i e (i) (n) c k (n)x(n k) Note that this is inconsistent with the notation used earlier, ŷ o (n) = M 1 h o (k)x(n k) = M 1 c k+1x(n k) Also the sums have M +1terms in them, rather than M terms as before Presumably motivated by a simple expression for the error e (i) (n) J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 3 J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 4

2 ˆx(n i) Types of Prediction c k(n)x(n k) k i e (i) (n) c k (n)x(n k) Forward Linear Prediction: i = Backward Linear Prediction: i = M Misnomer, but terminology is rooted in the literature Symmetric Linear Smoother: i = M/2 Linear Prediction Notation and Partitions x(n) [ x(n) x(n 1) x(n M +1) ] T x(n) [ x(n) x(n 1) x(n M) ] T = [ x(n) x T (n 1) ] T = [ x T (n) x(n M) ] T R(n) =E[x(n)x H (n)] R(n) =E[ x(n) x H (n)] Forward and backward linear prediction use specific partitions of the extended autocorrelation matrix [ Px (n) r R(n) f H(n) ] [ ] R(n) rb (n) R(n) r f (n) R(n 1) rb H(n) P x(n M) r f (n) =E[x(n 1)x (n)] r b (n) =E[x(n)x (n M)] J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 Forward and Backward Linear Prediction Estimator and Error ˆx f (n) = a k(n)x(n k) = a H (n)x(n 1) k=1 M 1 ˆx b (n) = b k(n)x(n k) e f = x(n)+ e b = M 1 = b H (n)x(n) a k(n)x(n k) = x(n)+a H (n)x(n 1) k=1 b k(n)x(n k)+x(n M) = b H (n)x(n)+x(n M) Again, new notation compared to the FIR linear estimation case I use subscripts for the f instead of superscripts like the text Forward and Backward Linear Prediction Solution Solution is the same as before, but watch the minus signs R(n 1)a o (n) = r f (n) R(n)b o (n) = r b (n) P f,o (n) =P x (n)+rf H (n)a o (n) P b,o (n) =P x (n M)+rb H (n)b o (n) ˆx f (n) = a k(n)x(n k) k=1 = a H (n)x(n 1) M 1 ˆx b (n) = b k(n)x(n k) = b H (n)x(n) J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 8

3 Stationary Case: Autocorrelation Matrix When the process is stationary, something surprising happens! R (M+1) (M+1) = 2 4 r x () r x (1) r x (M 1) r x (M) rx(1) r x () r x (M 2) r x (M 1) rx(m 1) rx(m 2) r x () r x (1) rx(m) rx(m 1) rx(1) r x () r T r x (1) r x (2) r x (M) Λ M 1 R(n) = [ Px (n) rf H(n) ] [ ] R(n) rb (n) = r f (n) R(n 1) rb H(n) P x(n M) 3 R (M+1) (M+1) = Stationary Case: Cross-correlation Vector 2 4 r x () r x (1) r x (M) rx(1) r x () r x (M 1) rx(m) rx(m 1) r x () [ ] rx () r R = f H r f R [ ] R rb R = rb H r x () 3 [ ] rx () r = T r R [ ] R Jr = r H J r x () r M r x (1) r x (2) r x (M) 3 Clearly, R(n) =R(n 1) P x () = P x (n M) =r x () r f =E[x(n 1)x (n)] r b =E[x(n)x (n M)] where J is the exchange matrix = r = Jr J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 9 J McNames Portland State University ECE 39/39 Linear Prediction Ver J 1 1 Exchange Matrix J H J = JJ H = I Counterpart to the identity matrix When multiplied on the left, flips a vector upside down When multiplied on the right, flips a vector sideways Don t do this in MATLAB many wasted multiplications by zeros See fliplr and flipud Forward/Backward Prediction Relationship Rb o = r b = Jr Ra o = r f = r JRa o = Jr JR a o = Jr = r b JR = RJ R(Ja o)= r b b o = Ja o The BLP parameter vector is the flipped and conjugated FLP parameter vector! Useful for estimation: can solve for both and combine them to reduce variance Further the prediction errors are the same! P f,o = P b,o = r() + r H a o = r() + r H Jb o J McNames Portland State University ECE 39/39 Linear Prediction Ver J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 12

4 Example 1: MA Process Create a synthetic MA process in MATLAB Plot the pole-zero and transfer function of the system Plot the MMSE versus the point being estimated, for M =2 ˆx(n i) c k(n)x(n k) k i MNMSE Example 1: MMSE Versus Prediction Index i Minimum NMSE Estimated MNMSE Prediction Index (i, Samples) J McNames Portland State University ECE 39/39 Linear Prediction Ver J McNames Portland State University ECE 39/39 Linear Prediction Ver Example 1: Prediction Example Example 1: Prediction Example M:2 i: NMSE:388 x(n) ˆx(n +) M:2 i:13 NMSE:18 x(n) ˆx(n +13) Signal + Estimate (scaled) Signal + Estimate (scaled) Sample Time (n) Sample Time (n) J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 1 J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 1

5 Signal + Estimate (scaled) Example 1: Prediction Example M:2 i:2 NMSE:388 x(n) ˆx(n +2) Sample Time (n) Example 1: MATLAB Code clear all; close all; N = ; % Number of samples M = 2; % Size of filter b = poly([99 98*j -98*j 98*exp(j*8*pi) 98*exp(-j*8*pi)]); nz = length(b)-1; % Number of zeros Mmax = M+1; % Maximum value to consider % Calculate the Auto- and Cross-Correlation rx = conv(b,fliplr(b)); % Quick calculation of rx k = -nz:nz; rx = rx(nz+1:end); % Trim off negative lags ryx = rx(2:end); % Cross-correlation one-step ahead % Generate Example w = randn(n,1); x = filter(b,1,w); % Build extended R Re = zeros(m+1,m+1); for c1=1:m+1, for c2=1:m+1, id = abs(c1-c2); if id<=nz, Re(c1,c2) = rx(id+1); % Locations of zeros J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 1 J McNames Portland State University ECE 39/39 Linear Prediction Ver Po = zeros(m+1,1); X = zeros(n,m+1); Poh = zeros(m+1,1); for id=:m, R = [Re(1:id,1:id),Re(1:id,id+2:end); Re(id+2:end,1:id),Re(id+2:end,id+2:end)]; d = [Re(1:id,id+1);Re(id+2:end,id+1)]; % Extract i th column sans the ith row Px = Re(id+1,id+1); co = -inv(r)*d; Po(id+1) = Px + d *co; X(:,id+1) = filter(-[co(1:id);;co(id+1:end)],1,x); k = 1:N-(id+1); Poh(id+1) = mean((x(k)-x(k+id,id+1))^2); % Plot MMSE Versus Order h1 = plot3(:m,poh/var(x),-1*ones(m+1,1), ro ); set(h1, MarkerFaceColor, r ); set(h1, MarkerSize,); view(,9); hold on; h2 = stem(:m,po/px, k ); set(h2, MarkerFaceColor, k ); set(h2, MarkerSize,4); hold off; xlabel( Prediction Index (i, Samples) ); ylabel( MNMSE ); xlim([- M+]); ylim([ 1]); AxisLines; legend([h1;h2], Minimum NMSE Estimated, MNMSE ); print( MAExampleMNMSEvi, -depsc ); % Plot Estimates id = [,round(m/2),m]; for c1=1:length(id), k = 1:N-(id(c1)+1); xh = X(k+id(c1),id(c1)+1); FigureSet(2, LTX ); h = plot(k,x(k), b,k,xh(k), g ); set(h, LineWidth,8); set(h, Marker, ); set(h, MarkerSize,8); xlim([ 1]); ylim(std(x)*[-3 3]); AxisLines; xlabel( Sample Time (n) ); ylabel( Signal + Estimate (scaled) ); set(get(gca, Title ), Interpreter, LaTeX ); title(sprintf( M:%d i:%d $\\widehat{\\mathrm{nmse}}$:%3f,m,id(c1),mean((x(k)-xh)^2)/var(x(k)))); hl = legend(h, $x(n)$,sprintf( $\\hat{x}(n+%d)$,id(c1))); set(hl, Interpreter, LaTeX ) print(sprintf( MAExamplePlot%2d,id(c1)), -depsc ); J McNames Portland State University ECE 39/39 Linear Prediction Ver J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 2

6 Selected Properties If x(n) is stationary, The linear smoother has linear phase The forward prediction error filter (PEF) is minimum phase The backward PEF is maximum phase The forward and backward prediction errors can be expressed as P f,o (n) = det R(n) det R(n 1) P b,o (n) = Other properties and proofs are given in the booik det R(n) det R(n) Example 2: Financial Time Series Forecast Try predicting a tremor signal acquired from an accelerometer attached to the wrist of a patient with Parkinson s disease J McNames Portland State University ECE 39/39 Linear Prediction Ver J McNames Portland State University ECE 39/39 Linear Prediction Ver Example 2: Data Details The files tremorparkgz and tremorphysgz contain measurements of the acceleration of the outstretched hand which is supported and fixed at the wrist The units are arbitrary The sampling frequency is 3 Hz tremorphysgz shows the tremor of a healthy person, tremorparkgz that of a person suffering from Parkinson s disease The real amplitude of the latter is of course much larger than that of the former For further information see: or mail to:jeti@fdmuni-freiburgde Jens Timmer Acceleration (scaled) N: Time (s) J McNames Portland State University ECE 39/39 Linear Prediction Ver J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 24

7 1 Autocorrelation 12 Partial Autocorrelation 1 8 ρ ρ Lag (s) Lag (s) J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 2 J McNames Portland State University ECE 39/39 Linear Prediction Ver x 1 Blackman-Tukey Estimated PSD PSD (scaled) 8 4 Frequency (Hz) Frequency (Hz) Signal Time (s) J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 2 J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 28

8 3 2 NMSE:42 P:1 M:2 Signal Predicted 1 8 P:1 Acceleration (scaled) NMSE Time (s) M (samples) J McNames Portland State University ECE 39/39 Linear Prediction Ver J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 3 NMSE M: P(s) Example 2: MATLAB Code clear; close all; % Load the Data x = load( R:/Tremor/Parkinsdat ); x = x - mean(x); fs = 3; % Sample rate (Hz) nx = length(x); % Length of data k = 1:nx; % Sample index t = (k-)/fs; % Sample times % Plot the Data h = plot(t(1:1:end),x(1:1:end), r ); set(h, LineWidth,8); xlabel( Time (s) ); ylabel( Acceleration (scaled) ); title(sprintf( N:%d,nx)); xlim([t(1) t(end)]); ylim(prctile(x,[ 99])); print( RESignal, -depsc ); % Plot the Autocorrelation Autocorrelation(x,fs,); J McNames Portland State University ECE 39/39 Linear Prediction Ver J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 32

9 print( REAutocorrelation, -depsc ); % Plot the Partial Autocorrelation PartialAutocorrelation(x,fs,); print( REPartialAutocorrelation, -depsc ); % Plot the Power Spectral Density BlackmanTukey(x,fs,1); print( REBlackmanTukey, -depsc ); % Plot the Spectrogram NonparametricSpectrogram(decimate(x,),fs/,2); print( RESpectrogram, -depsc ); % Calculate the Auto- and Cross-Correlation rx = Autocorrelation(x,fs,2); m = 2; % Filter length p = 1; % Number of steps ahead to predict R = zeros(m,m); for c2=1:m, R(c1,c2) = rx(abs(c1-c2)+1); d = zeros(m,1); d(c1) = rx(c1+p); co = inv(r)*d; xh = filter(co,1,[zeros(p,1);x]); xh = xh(1:nx); NMSE = mean((x-xh)^2)/mean(x^2); % Plot Segment of Signal and Predicted h = plot(t,x, r,t,xh, g ); set(h(1), LineWidth,8); set(h(2), LineWidth,12); xlabel( Time (s) ); ylabel( Acceleration (scaled) ); title(sprintf( NMSE:%3f P:%d M:%d,NMSE,p,m)); legend(h, Signal, Predicted ); xlim([t(1) ]); ylim(prctile(x,[ 99])); print( RESignalPredicted, -depsc ); % Sweep M for P=1 J McNames Portland State University ECE 39/39 Linear Prediction Ver J McNames Portland State University ECE 39/39 Linear Prediction Ver p = 1; % Number of steps ahead to predict M = 1:1; nm = length(m); NMSE = zeros(nm,1); for c=1:nm, m = M(c); % Filter length R = zeros(m,m); for c2=1:m, R(c1,c2) = rx(abs(c1-c2)+1); d = zeros(m,1); d(c1) = rx(c1+p); co = inv(r)*d; xh = filter(co,1,[zeros(p,1);x]); xh = xh(1:nx); NMSE(c) = mean((x-xh)^2)/mean(x^2); h = plot(m,nmse, b ); set(h(1), LineWidth,8); xlabel( M (samples) ); ylabel( NMSE ); title(sprintf( P:%d,p)); xlim([ M(end)+]); ylim([ 1]); print( RESweepM, -depsc ); % Sweep P for P=1 P = 1:3; % Number of steps ahead to predict m = 2; np = length(p); NMSE = zeros(np,1); R = zeros(m,m); for c2=1:m, R(c1,c2) = rx(abs(c1-c2)+1); for c=1:np, p = P(c); % Filter length d = zeros(m,1); d(c1) = rx(c1+p); co = inv(r)*d; xh = filter(co,1,[zeros(p,1);x]); xh = xh(1:nx); NMSE(c) = mean((x-xh)^2)/mean(x^2); h = plot(p/fs,nmse, b ); set(h, LineWidth,8); set(h, Marker, ); set(h, MarkerSize,); hold on; hb = plot([ P(end)/fs],[1 1], r: ); hold off; J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 3 J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 3

10 xlabel( P (s) ); ylabel( NMSE ); title(sprintf( M:%d,m)); xlim([ P(end)/fs]); ylim([ 1]); print( RESweepP, -depsc ); J McNames Portland State University ECE 39/39 Linear Prediction Ver 12 3

Fast Fourier Transform Discrete-time windowing Discrete Fourier Transform Relationship to DTFT Relationship to DTFS Zero padding

Fast Fourier Transform Discrete-time windowing Discrete Fourier Transform Relationship to DTFT Relationship to DTFS Zero padding Fast Fourier Transform Discrete-time windowing Discrete Fourier Transform Relationship to DTFT Relationship to DTFS Zero padding J. McNames Portland State University ECE 223 FFT Ver. 1.03 1 Fourier Series

More information

Fast Fourier Transform Discrete-time windowing Discrete Fourier Transform Relationship to DTFT Relationship to DTFS Zero padding

Fast Fourier Transform Discrete-time windowing Discrete Fourier Transform Relationship to DTFT Relationship to DTFS Zero padding Fast Fourier Transform Discrete-time windowing Discrete Fourier Transform Relationship to DTFT Relationship to DTFS Zero padding Fourier Series & Transform Summary x[n] = X[k] = 1 N k= n= X[k]e jkω

More information

J. McNames Portland State University ECE 223 DT Fourier Series Ver

J. McNames Portland State University ECE 223 DT Fourier Series Ver Overview of DT Fourier Series Topics Orthogonality of DT exponential harmonics DT Fourier Series as a Design Task Picking the frequencies Picking the range Finding the coefficients Example J. McNames Portland

More information

Overview of Discrete-Time Fourier Transform Topics Handy Equations Handy Limits Orthogonality Defined orthogonal

Overview of Discrete-Time Fourier Transform Topics Handy Equations Handy Limits Orthogonality Defined orthogonal Overview of Discrete-Time Fourier Transform Topics Handy equations and its Definition Low- and high- discrete-time frequencies Convergence issues DTFT of complex and real sinusoids Relationship to LTI

More information

Order Recursion Introduction Order versus Time Updates Matrix Inversion by Partitioning Lemma Levinson Algorithm Interpretations Examples

Order Recursion Introduction Order versus Time Updates Matrix Inversion by Partitioning Lemma Levinson Algorithm Interpretations Examples Order Recursion Introduction Order versus Tie Updates Matrix Inversion by Partitioning Lea Levinson Algorith Interpretations Exaples Introduction Rc d There are any ways to solve the noral equations Solutions

More information

Spectrograms Overview Introduction and motivation Windowed estimates as filter banks Uncertainty principle Examples Other estimates

Spectrograms Overview Introduction and motivation Windowed estimates as filter banks Uncertainty principle Examples Other estimates Spectrograms Overview Introduction and motivation Windowed estimates as filter banks Uncertainty principle Examples Other estimates J. McNames Portland State University ECE 3 Spectrograms Ver. 1.1 1 Introduction

More information

x n C l 1 u n C m 1 y n C p 1 v n C p 1

x n C l 1 u n C m 1 y n C p 1 v n C p 1 Kalman Filter Derivation Examples Time and Measurement Updates x u n v n, State Space Model Review x n+ = F n x n + G n u n x u k v k y n = H n x n + v n Π = Q n δ nk S n δ nk Snδ nk R n δ nk F n C l l

More information

Core Concepts Review. Orthogonality of Complex Sinusoids Consider two (possibly non-harmonic) complex sinusoids

Core Concepts Review. Orthogonality of Complex Sinusoids Consider two (possibly non-harmonic) complex sinusoids Overview of Continuous-Time Fourier Transform Topics Definition Compare & contrast with Laplace transform Conditions for existence Relationship to LTI systems Examples Ideal lowpass filters Relationship

More information

Probability Space. J. McNames Portland State University ECE 538/638 Stochastic Signals Ver

Probability Space. J. McNames Portland State University ECE 538/638 Stochastic Signals Ver Stochastic Signals Overview Definitions Second order statistics Stationarity and ergodicity Random signal variability Power spectral density Linear systems with stationary inputs Random signal memory Correlation

More information

Extended Kalman Filter Derivation Example application: frequency tracking

Extended Kalman Filter Derivation Example application: frequency tracking Extended Kalman Filter Derivation Example application: frequency tracking Nonlinear State Space Model Consider the nonlinear state-space model of a random process x n+ = f n (x n )+g n (x n )u n y n =

More information

Introduction. Spectral Estimation Overview Periodogram Bias, variance, and distribution Blackman-Tukey Method Welch-Bartlett Method Others

Introduction. Spectral Estimation Overview Periodogram Bias, variance, and distribution Blackman-Tukey Method Welch-Bartlett Method Others Spectral Estimation Overview Periodogram Bias, variance, and distribution Blackman-Tukey Method Welch-Bartlett Method Others Introduction R x (e jω ) r x (l)e jωl l= Most stationary random processes have

More information

Windowing Overview Relevance Main lobe/side lobe tradeoff Popular windows Examples

Windowing Overview Relevance Main lobe/side lobe tradeoff Popular windows Examples ing Overview Relevance Main lobe/side lobe tradeoff Popular windows Examples ing In practice, we cannot observe a signal x(n) for n = to n = Thus we must truncate, or window, the signal: x(n) w(n) ing

More information

J. McNames Portland State University ECE 223 Sampling Ver

J. McNames Portland State University ECE 223 Sampling Ver Overview of Sampling Topics (Shannon) sampling theorem Impulse-train sampling Interpolation (continuous-time signal reconstruction) Aliasing Relationship of CTFT to DTFT DT processing of CT signals DT

More information

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1

Terminology Suppose we have N observations {x(n)} N 1. Estimators as Random Variables. {x(n)} N 1 Estimation Theory Overview Properties Bias, Variance, and Mean Square Error Cramér-Rao lower bound Maximum likelihood Consistency Confidence intervals Properties of the mean estimator Properties of the

More information

Overview of Sampling Topics

Overview of Sampling Topics Overview of Sampling Topics (Shannon) sampling theorem Impulse-train sampling Interpolation (continuous-time signal reconstruction) Aliasing Relationship of CTFT to DTFT DT processing of CT signals DT

More information

Nonparametric and Parametric Defined This text distinguishes between systems and the sequences (processes) that result when a WN input is applied

Nonparametric and Parametric Defined This text distinguishes between systems and the sequences (processes) that result when a WN input is applied Linear Signal Models Overview Introduction Linear nonparametric vs. parametric models Equivalent representations Spectral flatness measure PZ vs. ARMA models Wold decomposition Introduction Many researchers

More information

x(n) = a k x(n k)+w(n) k=1

x(n) = a k x(n k)+w(n) k=1 Autoregressive Models Overview Direct structures Types of estimators Parametric spectral estimation Parametric time-frequency analysis Order selection criteria Lattice structures? Maximum entropy Excitation

More information

Solutions. Chapter 5. Problem 5.1. Solution. Consider the driven, two-well Duffing s oscillator. which can be written in state variable form as

Solutions. Chapter 5. Problem 5.1. Solution. Consider the driven, two-well Duffing s oscillator. which can be written in state variable form as Chapter 5 Solutions Problem 5.1 Consider the driven, two-well Duffing s oscillator which can be written in state variable form as ẍ + ɛγẋ x + x 3 = ɛf cos(ωt) ẋ = v v = x x 3 + ɛ( γv + F cos(ωt)). In the

More information

Assignment 2 Solutions Fourier Series

Assignment 2 Solutions Fourier Series Assignment 2 Solutions Fourier Series ECE 223 Signals and Systems II Version.2 Spring 26. DT Fourier Series. Consider the following discrete-time periodic signal n = +7l n =+7l x[n] = n =+7l Otherwise

More information

Overview of Bode Plots Transfer function review Piece-wise linear approximations First-order terms Second-order terms (complex poles & zeros)

Overview of Bode Plots Transfer function review Piece-wise linear approximations First-order terms Second-order terms (complex poles & zeros) Overview of Bode Plots Transfer function review Piece-wise linear approximations First-order terms Second-order terms (complex poles & zeros) J. McNames Portland State University ECE 222 Bode Plots Ver.

More information

Practical Spectral Estimation

Practical Spectral Estimation Digital Signal Processing/F.G. Meyer Lecture 4 Copyright 2015 François G. Meyer. All Rights Reserved. Practical Spectral Estimation 1 Introduction The goal of spectral estimation is to estimate how the

More information

Random signals II. ÚPGM FIT VUT Brno,

Random signals II. ÚPGM FIT VUT Brno, Random signals II. Jan Černocký ÚPGM FIT VUT Brno, cernocky@fit.vutbr.cz 1 Temporal estimate of autocorrelation coefficients for ergodic discrete-time random process. ˆR[k] = 1 N N 1 n=0 x[n]x[n + k],

More information

Least Square Es?ma?on, Filtering, and Predic?on: ECE 5/639 Sta?s?cal Signal Processing II: Linear Es?ma?on

Least Square Es?ma?on, Filtering, and Predic?on: ECE 5/639 Sta?s?cal Signal Processing II: Linear Es?ma?on Least Square Es?ma?on, Filtering, and Predic?on: Sta?s?cal Signal Processing II: Linear Es?ma?on Eric Wan, Ph.D. Fall 2015 1 Mo?va?ons If the second-order sta?s?cs are known, the op?mum es?mator is given

More information

Laboratory Project 1: Introduction to Random Processes

Laboratory Project 1: Introduction to Random Processes Laboratory Project 1: Introduction to Random Processes Random Processes With Applications (MVE 135) Mats Viberg Department of Signals and Systems Chalmers University of Technology 412 96 Gteborg, Sweden

More information

Lecture 7: Linear Prediction

Lecture 7: Linear Prediction 1 Lecture 7: Linear Prediction Overview Dealing with three notions: PREDICTION, PREDICTOR, PREDICTION ERROR; FORWARD versus BACKWARD: Predicting the future versus (improper terminology) predicting the

More information

Assignment 4 Solutions Continuous-Time Fourier Transform

Assignment 4 Solutions Continuous-Time Fourier Transform Assignment 4 Solutions Continuous-Time Fourier Transform ECE 3 Signals and Systems II Version 1.01 Spring 006 1. Properties of complex numbers. Let c 1 α 1 + jβ 1 and c α + jβ be two complex numbers. a.

More information

Department of Electrical and Computer Engineering Digital Speech Processing Homework No. 6 Solutions

Department of Electrical and Computer Engineering Digital Speech Processing Homework No. 6 Solutions Problem 1 Department of Electrical and Computer Engineering Digital Speech Processing Homework No. 6 Solutions The complex cepstrum, ˆx[n], of a sequence x[n] is the inverse Fourier transform of the complex

More information

Chapter 2 Wiener Filtering

Chapter 2 Wiener Filtering Chapter 2 Wiener Filtering Abstract Before moving to the actual adaptive filtering problem, we need to solve the optimum linear filtering problem (particularly, in the mean-square-error sense). We start

More information

EE538 Final Exam Fall 2007 Mon, Dec 10, 8-10 am RHPH 127 Dec. 10, Cover Sheet

EE538 Final Exam Fall 2007 Mon, Dec 10, 8-10 am RHPH 127 Dec. 10, Cover Sheet EE538 Final Exam Fall 2007 Mon, Dec 10, 8-10 am RHPH 127 Dec. 10, 2007 Cover Sheet Test Duration: 120 minutes. Open Book but Closed Notes. Calculators allowed!! This test contains five problems. Each of

More information

Solutions for examination in TSRT78 Digital Signal Processing,

Solutions for examination in TSRT78 Digital Signal Processing, Solutions for examination in TSRT78 Digital Signal Processing, 2014-04-14 1. s(t) is generated by s(t) = 1 w(t), 1 + 0.3q 1 Var(w(t)) = σ 2 w = 2. It is measured as y(t) = s(t) + n(t) where n(t) is white

More information

Linear Stochastic Models. Special Types of Random Processes: AR, MA, and ARMA. Digital Signal Processing

Linear Stochastic Models. Special Types of Random Processes: AR, MA, and ARMA. Digital Signal Processing Linear Stochastic Models Special Types of Random Processes: AR, MA, and ARMA Digital Signal Processing Department of Electrical and Electronic Engineering, Imperial College d.mandic@imperial.ac.uk c Danilo

More information

Statistical methods. Mean value and standard deviations Standard statistical distributions Linear systems Matrix algebra

Statistical methods. Mean value and standard deviations Standard statistical distributions Linear systems Matrix algebra Statistical methods Mean value and standard deviations Standard statistical distributions Linear systems Matrix algebra Statistical methods Generating random numbers MATLAB has many built-in functions

More information

Statistical and Adaptive Signal Processing

Statistical and Adaptive Signal Processing r Statistical and Adaptive Signal Processing Spectral Estimation, Signal Modeling, Adaptive Filtering and Array Processing Dimitris G. Manolakis Massachusetts Institute of Technology Lincoln Laboratory

More information

ADSP ADSP ADSP ADSP. Advanced Digital Signal Processing (18-792) Spring Fall Semester, Department of Electrical and Computer Engineering

ADSP ADSP ADSP ADSP. Advanced Digital Signal Processing (18-792) Spring Fall Semester, Department of Electrical and Computer Engineering Advanced Digital Signal rocessing (18-792) Spring Fall Semester, 201 2012 Department of Electrical and Computer Engineering ROBLEM SET 8 Issued: 10/26/18 Due: 11/2/18 Note: This problem set is due Friday,

More information

Bayesian Analysis - A First Example

Bayesian Analysis - A First Example Bayesian Analysis - A First Example This script works through the example in Hoff (29), section 1.2.1 We are interested in a single parameter: θ, the fraction of individuals in a city population with with

More information

Basics on 2-D 2 D Random Signal

Basics on 2-D 2 D Random Signal Basics on -D D Random Signal Spring 06 Instructor: K. J. Ray Liu ECE Department, Univ. of Maryland, College Park Overview Last Time: Fourier Analysis for -D signals Image enhancement via spatial filtering

More information

ECE531 Lecture 11: Dynamic Parameter Estimation: Kalman-Bucy Filter

ECE531 Lecture 11: Dynamic Parameter Estimation: Kalman-Bucy Filter ECE531 Lecture 11: Dynamic Parameter Estimation: Kalman-Bucy Filter D. Richard Brown III Worcester Polytechnic Institute 09-Apr-2009 Worcester Polytechnic Institute D. Richard Brown III 09-Apr-2009 1 /

More information

Adaptive Systems Homework Assignment 1

Adaptive Systems Homework Assignment 1 Signal Processing and Speech Communication Lab. Graz University of Technology Adaptive Systems Homework Assignment 1 Name(s) Matr.No(s). The analytical part of your homework (your calculation sheets) as

More information

Parametric Method Based PSD Estimation using Gaussian Window

Parametric Method Based PSD Estimation using Gaussian Window International Journal of Engineering Trends and Technology (IJETT) Volume 29 Number 1 - November 215 Parametric Method Based PSD Estimation using Gaussian Window Pragati Sheel 1, Dr. Rajesh Mehra 2, Preeti

More information

Laboratory Project 2: Spectral Analysis and Optimal Filtering

Laboratory Project 2: Spectral Analysis and Optimal Filtering Laboratory Project 2: Spectral Analysis and Optimal Filtering Random signals analysis (MVE136) Mats Viberg and Lennart Svensson Department of Signals and Systems Chalmers University of Technology 412 96

More information

10 Linear Prediction. Statistics 626

10 Linear Prediction. Statistics 626 10 Linear Prediction If we have a realization x(1),...,x(n) fromatimeseriesx, we would like a rule for how to take the x s and the probability properties of X to find the function of the data that best

More information

ELEG-636: Statistical Signal Processing

ELEG-636: Statistical Signal Processing ELEG-636: Statistical Signal Processing Gonzalo R. Arce Department of Electrical and Computer Engineering University of Delaware Spring 2010 Gonzalo R. Arce (ECE, Univ. of Delaware) ELEG-636: Statistical

More information

EE5585 Data Compression April 18, Lecture 23

EE5585 Data Compression April 18, Lecture 23 EE5585 Data Compression April 18, 013 Lecture 3 Instructor: Arya Mazumdar Scribe: Trevor Webster Differential Encoding Suppose we have a signal that is slowly varying For instance, if we were looking at

More information

5 Kalman filters. 5.1 Scalar Kalman filter. Unit delay Signal model. System model

5 Kalman filters. 5.1 Scalar Kalman filter. Unit delay Signal model. System model 5 Kalman filters 5.1 Scalar Kalman filter 5.1.1 Signal model System model {Y (n)} is an unobservable sequence which is described by the following state or system equation: Y (n) = h(n)y (n 1) + Z(n), n

More information

Advanced Digital Signal Processing -Introduction

Advanced Digital Signal Processing -Introduction Advanced Digital Signal Processing -Introduction LECTURE-2 1 AP9211- ADVANCED DIGITAL SIGNAL PROCESSING UNIT I DISCRETE RANDOM SIGNAL PROCESSING Discrete Random Processes- Ensemble Averages, Stationary

More information

Lab 4: Quantization, Oversampling, and Noise Shaping

Lab 4: Quantization, Oversampling, and Noise Shaping Lab 4: Quantization, Oversampling, and Noise Shaping Due Friday 04/21/17 Overview: This assignment should be completed with your assigned lab partner(s). Each group must turn in a report composed using

More information

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d)

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d) Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes (cont d) Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides

More information

Discrete-Time Systems

Discrete-Time Systems FIR Filters With this chapter we turn to systems as opposed to signals. The systems discussed in this chapter are finite impulse response (FIR) digital filters. The term digital filter arises because these

More information

Math 515 Fall, 2008 Homework 2, due Friday, September 26.

Math 515 Fall, 2008 Homework 2, due Friday, September 26. Math 515 Fall, 2008 Homework 2, due Friday, September 26 In this assignment you will write efficient MATLAB codes to solve least squares problems involving block structured matrices known as Kronecker

More information

LAB 2: DTFT, DFT, and DFT Spectral Analysis Summer 2011

LAB 2: DTFT, DFT, and DFT Spectral Analysis Summer 2011 University of Illinois at Urbana-Champaign Department of Electrical and Computer Engineering ECE 311: Digital Signal Processing Lab Chandra Radhakrishnan Peter Kairouz LAB 2: DTFT, DFT, and DFT Spectral

More information

EE364a Homework 8 solutions

EE364a Homework 8 solutions EE364a, Winter 2007-08 Prof. S. Boyd EE364a Homework 8 solutions 9.8 Steepest descent method in l -norm. Explain how to find a steepest descent direction in the l -norm, and give a simple interpretation.

More information

Temperature measurement

Temperature measurement Luleå University of Technology Johan Carlson Last revision: July 22, 2009 Measurement Technology and Uncertainty Analysis - E7021E Lab 3 Temperature measurement Introduction In this lab you are given a

More information

ECE 541 Stochastic Signals and Systems Problem Set 9 Solutions

ECE 541 Stochastic Signals and Systems Problem Set 9 Solutions ECE 541 Stochastic Signals and Systems Problem Set 9 Solutions Problem Solutions : Yates and Goodman, 9.5.3 9.1.4 9.2.2 9.2.6 9.3.2 9.4.2 9.4.6 9.4.7 and Problem 9.1.4 Solution The joint PDF of X and Y

More information

The Z-Transform. For a phasor: X(k) = e jωk. We have previously derived: Y = H(z)X

The Z-Transform. For a phasor: X(k) = e jωk. We have previously derived: Y = H(z)X The Z-Transform For a phasor: X(k) = e jωk We have previously derived: Y = H(z)X That is, the output of the filter (Y(k)) is derived by multiplying the input signal (X(k)) by the transfer function (H(z)).

More information

Homework 1 Solutions

Homework 1 Solutions 18-9 Signals and Systems Profs. Byron Yu and Pulkit Grover Fall 18 Homework 1 Solutions Part One 1. (8 points) Consider the DT signal given by the algorithm: x[] = 1 x[1] = x[n] = x[n 1] x[n ] (a) Plot

More information

ENGR352 Problem Set 02

ENGR352 Problem Set 02 engr352/engr352p02 September 13, 2018) ENGR352 Problem Set 02 Transfer function of an estimator 1. Using Eq. (1.1.4-27) from the text, find the correct value of r ss (the result given in the text is incorrect).

More information

1. Determine if each of the following are valid autocorrelation matrices of WSS processes. (Correlation Matrix),R c =

1. Determine if each of the following are valid autocorrelation matrices of WSS processes. (Correlation Matrix),R c = ENEE630 ADSP Part II w/ solution. Determine if each of the following are valid autocorrelation matrices of WSS processes. (Correlation Matrix) R a = 4 4 4,R b = 0 0,R c = j 0 j 0 j 0 j 0 j,r d = 0 0 0

More information

Notes perso. - cb1 1. Version 3 - November u(t)e i2πft dt u(t) =

Notes perso. - cb1 1. Version 3 - November u(t)e i2πft dt u(t) = Notes perso. - cb1 1 Tutorial : fft, psd & coherence with Matlab Version 3 - November 01 1 Fourier transform by Fast Fourier Transform (FFT) Definition of the Fourier transform : û(f) = F[u(t)] = As an

More information

Errata and Notes for

Errata and Notes for Errata and Notes for Statistical and Adaptive Signal Processing Dimitris G. Manolakis, Vinay K. Ingle, and Stephen M. Kogon, Artech Housre, Inc., c 2005, ISBN 1580536107. James McNames March 15, 2006 Errata

More information

Least Squares with Examples in Signal Processing 1. 2 Overdetermined equations. 1 Notation. The sum of squares of x is denoted by x 2 2, i.e.

Least Squares with Examples in Signal Processing 1. 2 Overdetermined equations. 1 Notation. The sum of squares of x is denoted by x 2 2, i.e. Least Squares with Eamples in Signal Processing Ivan Selesnick March 7, 3 NYU-Poly These notes address (approimate) solutions to linear equations by least squares We deal with the easy case wherein the

More information

DFT-Based FIR Filtering. See Porat s Book: 4.7, 5.6

DFT-Based FIR Filtering. See Porat s Book: 4.7, 5.6 DFT-Based FIR Filtering See Porat s Book: 4.7, 5.6 1 Motivation: DTFT View of Filtering There are two views of filtering: * Time Domain * Frequency Domain x[ X f ( θ ) h[ H f ( θ ) Y y[ = h[ * x[ f ( θ

More information

ECE 541 Stochastic Signals and Systems Problem Set 11 Solution

ECE 541 Stochastic Signals and Systems Problem Set 11 Solution ECE 54 Stochastic Signals and Systems Problem Set Solution Problem Solutions : Yates and Goodman,..4..7.3.3.4.3.8.3 and.8.0 Problem..4 Solution Since E[Y (t] R Y (0, we use Theorem.(a to evaluate R Y (τ

More information

w n = c k v n k (1.226) w n = c k v n k + d k w n k (1.227) Clearly non-recursive filters are a special case of recursive filters where M=0.

w n = c k v n k (1.226) w n = c k v n k + d k w n k (1.227) Clearly non-recursive filters are a special case of recursive filters where M=0. Random Data 79 1.13 Digital Filters There are two fundamental types of digital filters Non-recursive N w n = c k v n k (1.226) k= N and recursive N M w n = c k v n k + d k w n k (1.227) k= N k=1 Clearly

More information

Stationary or Non-Stationary Random Excitation for Vibration-Based Structural Damage Detection? An exploratory study

Stationary or Non-Stationary Random Excitation for Vibration-Based Structural Damage Detection? An exploratory study Stationary or Non-Stationary Random Excitation for Vibration-Based Structural Damage Detection? An exploratory study Andriana S. GEORGANTOPOULOU & Spilios D. FASSOIS Stochastic Mechanical Systems & Automation

More information

Complement on Digital Spectral Analysis and Optimal Filtering: Theory and Exercises

Complement on Digital Spectral Analysis and Optimal Filtering: Theory and Exercises Complement on Digital Spectral Analysis and Optimal Filtering: Theory and Exercises Random Processes With Applications (MVE 135) Mats Viberg Department of Signals and Systems Chalmers University of Technology

More information

Centre for Mathematical Sciences HT 2017 Mathematical Statistics. Study chapters 6.1, 6.2 and in the course book.

Centre for Mathematical Sciences HT 2017 Mathematical Statistics. Study chapters 6.1, 6.2 and in the course book. Lund University Stationary stochastic processes Centre for Mathematical Sciences HT 2017 Mathematical Statistics Computer exercise 2 in Stationary stochastic processes, HT 17. The purpose with this computer

More information

Naive Bayes & Introduction to Gaussians

Naive Bayes & Introduction to Gaussians Naive Bayes & Introduction to Gaussians Andreas C. Kapourani 2 March 217 1 Naive Bayes classifier In the previous lab we illustrated how to use Bayes Theorem for pattern classification, which in practice

More information

Lecture 7 Random Signal Analysis

Lecture 7 Random Signal Analysis Lecture 7 Random Signal Analysis 7. Introduction to Probability 7. Amplitude Distributions 7.3 Uniform, Gaussian, and Other Distributions 7.4 Power and Power Density Spectra 7.5 Properties of the Power

More information

EE266 Homework 3 Solutions

EE266 Homework 3 Solutions EE266, Spring 2014-15 Professor S. Lall EE266 Homework 3 Solutions 1. Second passage time. In this problem we will consider the following Markov chain. Note that self-loops are omitted from this figure.

More information

Lecture - 30 Stationary Processes

Lecture - 30 Stationary Processes Probability and Random Variables Prof. M. Chakraborty Department of Electronics and Electrical Communication Engineering Indian Institute of Technology, Kharagpur Lecture - 30 Stationary Processes So,

More information

ELEG 833. Nonlinear Signal Processing

ELEG 833. Nonlinear Signal Processing Nonlinear Signal Processing ELEG 833 Gonzalo R. Arce Department of Electrical and Computer Engineering University of Delaware arce@ee.udel.edu February 15, 2005 1 INTRODUCTION 1 Introduction Signal processing

More information

CHAPTER 2. Frequency Domain Analysis

CHAPTER 2. Frequency Domain Analysis FREQUENCY DOMAIN ANALYSIS 16 CHAPTER 2 Frequency Domain Analysis ASSESSMENTOF FREQUENCY DOMAIN FORCE IDENTIFICATION PROCEDURES CHAPTE,R 2. FREQUENCY DOMAINANALYSIS 17 2. FREQUENCY DOMAIN ANALYSIS The force

More information

Complement on Digital Spectral Analysis and Optimal Filtering: Theory and Exercises

Complement on Digital Spectral Analysis and Optimal Filtering: Theory and Exercises Complement on Digital Spectral Analysis and Optimal Filtering: Theory and Exercises Random Signals Analysis (MVE136) Mats Viberg Department of Signals and Systems Chalmers University of Technology 412

More information

Topic # Feedback Control Systems

Topic # Feedback Control Systems Topic #16 16.31 Feedback Control Systems State-Space Systems Closed-loop control using estimators and regulators. Dynamics output feedback Back to reality Fall 2007 16.31 16 1 Combined Estimators and Regulators

More information

Test 2 Electrical Engineering Bachelor Module 8 Signal Processing and Communications

Test 2 Electrical Engineering Bachelor Module 8 Signal Processing and Communications Test 2 Electrical Engineering Bachelor Module 8 Signal Processing and Communications (201400432) Tuesday May 26, 2015, 14:00-17:00h This test consists of three parts, corresponding to the three courses

More information

to have roots with negative real parts, the necessary and sufficient conditions are that:

to have roots with negative real parts, the necessary and sufficient conditions are that: THE UNIVERSITY OF TEXAS AT SAN ANTONIO EE 543 LINEAR SYSTEMS AND CONTROL H O M E W O R K # 7 Sebastian A. Nugroho November 6, 7 Due date of the homework is: Sunday, November 6th @ :59pm.. The following

More information

3.0 PROBABILITY, RANDOM VARIABLES AND RANDOM PROCESSES

3.0 PROBABILITY, RANDOM VARIABLES AND RANDOM PROCESSES 3.0 PROBABILITY, RANDOM VARIABLES AND RANDOM PROCESSES 3.1 Introduction In this chapter we will review the concepts of probabilit, rom variables rom processes. We begin b reviewing some of the definitions

More information

EEO 401 Digital Signal Processing Prof. Mark Fowler

EEO 401 Digital Signal Processing Prof. Mark Fowler EEO 401 Digital Signal Processing Prof. Mark Fowler Note Set #21 Using the DFT to Implement FIR Filters Reading Assignment: Sect. 7.3 of Proakis & Manolakis Motivation: DTFT View of Filtering There are

More information

Tu: 9/3/13 Math 471, Fall 2013, Section 001 Lecture 1

Tu: 9/3/13 Math 471, Fall 2013, Section 001 Lecture 1 Tu: 9/3/13 Math 71, Fall 2013, Section 001 Lecture 1 1 Course intro Notes : Take attendance. Instructor introduction. Handout : Course description. Note the exam days (and don t be absent). Bookmark the

More information

Time Series Analysis

Time Series Analysis Time Series Analysis hm@imm.dtu.dk Informatics and Mathematical Modelling Technical University of Denmark DK-2800 Kgs. Lyngby 1 Outline of the lecture Input-Output systems The z-transform important issues

More information

Section 4.3. Polynomial Division; The Remainder Theorem and the Factor Theorem

Section 4.3. Polynomial Division; The Remainder Theorem and the Factor Theorem Section 4.3 Polynomial Division; The Remainder Theorem and the Factor Theorem Polynomial Long Division Let s compute 823 5 : Example of Long Division of Numbers Example of Long Division of Numbers Let

More information

Testing the MC Model. 9.1 Introduction. 9.2 The Size and Solution of the MC model

Testing the MC Model. 9.1 Introduction. 9.2 The Size and Solution of the MC model 9 Testing the MC Model 9.1 Introduction This chapter is concerned with testing the overall MC model. It is the counterpart of Chapter 8 for the US model. There are, however, many fewer tests in this chapter

More information

EE538 Final Exam Fall :20 pm -5:20 pm PHYS 223 Dec. 17, Cover Sheet

EE538 Final Exam Fall :20 pm -5:20 pm PHYS 223 Dec. 17, Cover Sheet EE538 Final Exam Fall 005 3:0 pm -5:0 pm PHYS 3 Dec. 17, 005 Cover Sheet Test Duration: 10 minutes. Open Book but Closed Notes. Calculators ARE allowed!! This test contains five problems. Each of the five

More information

Generalized Sidelobe Canceller and MVDR Power Spectrum Estimation. Bhaskar D Rao University of California, San Diego

Generalized Sidelobe Canceller and MVDR Power Spectrum Estimation. Bhaskar D Rao University of California, San Diego Generalized Sidelobe Canceller and MVDR Power Spectrum Estimation Bhaskar D Rao University of California, San Diego Email: brao@ucsd.edu Reference Books 1. Optimum Array Processing, H. L. Van Trees 2.

More information

Solutions. Chapter 1. Problem 1.1. Solution. Simulate free response of damped harmonic oscillator

Solutions. Chapter 1. Problem 1.1. Solution. Simulate free response of damped harmonic oscillator Chapter Solutions Problem. Simulate free response of damped harmonic oscillator ẍ + ζẋ + x = for different values of damping ration ζ and initial conditions. Plot the response x(t) and ẋ(t) versus t and

More information

Time Series Prediction by Kalman Smoother with Cross-Validated Noise Density

Time Series Prediction by Kalman Smoother with Cross-Validated Noise Density Time Series Prediction by Kalman Smoother with Cross-Validated Noise Density Simo Särkkä E-mail: simo.sarkka@hut.fi Aki Vehtari E-mail: aki.vehtari@hut.fi Jouko Lampinen E-mail: jouko.lampinen@hut.fi Abstract

More information

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes

Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes Parametric Signal Modeling and Linear Prediction Theory 1. Discrete-time Stochastic Processes Electrical & Computer Engineering North Carolina State University Acknowledgment: ECE792-41 slides were adapted

More information

CHAPTER 2 RANDOM PROCESSES IN DISCRETE TIME

CHAPTER 2 RANDOM PROCESSES IN DISCRETE TIME CHAPTER 2 RANDOM PROCESSES IN DISCRETE TIME Shri Mata Vaishno Devi University, (SMVDU), 2013 Page 13 CHAPTER 2 RANDOM PROCESSES IN DISCRETE TIME When characterizing or modeling a random variable, estimates

More information

MATRICES. a m,1 a m,n A =

MATRICES. a m,1 a m,n A = MATRICES Matrices are rectangular arrays of real or complex numbers With them, we define arithmetic operations that are generalizations of those for real and complex numbers The general form a matrix of

More information

(Extended) Kalman Filter

(Extended) Kalman Filter (Extended) Kalman Filter Brian Hunt 7 June 2013 Goals of Data Assimilation (DA) Estimate the state of a system based on both current and all past observations of the system, using a model for the system

More information

STA 4273H: Statistical Machine Learning

STA 4273H: Statistical Machine Learning STA 4273H: Statistical Machine Learning Russ Salakhutdinov Department of Statistics! rsalakhu@utstat.toronto.edu! http://www.utstat.utoronto.ca/~rsalakhu/ Sidney Smith Hall, Room 6002 Lecture 11 Project

More information

Correlation, discrete Fourier transforms and the power spectral density

Correlation, discrete Fourier transforms and the power spectral density Correlation, discrete Fourier transforms and the power spectral density visuals to accompany lectures, notes and m-files by Tak Igusa tigusa@jhu.edu Department of Civil Engineering Johns Hopkins University

More information

ECE 636: Systems identification

ECE 636: Systems identification ECE 636: Systems identification Lectures 3 4 Random variables/signals (continued) Random/stochastic vectors Random signals and linear systems Random signals in the frequency domain υ ε x S z + y Experimental

More information

MATH 552 Spectral Methods Spring Homework Set 5 - SOLUTIONS

MATH 552 Spectral Methods Spring Homework Set 5 - SOLUTIONS MATH 55 Spectral Methods Spring 9 Homework Set 5 - SOLUTIONS. Suppose you are given an n n linear system Ax = f where the matrix A is tridiagonal b c a b c. A =.........,. a n b n c n a n b n with x =

More information

0 for all other times. Suppose further that we sample the signal with a sampling frequency f ;. We can write x # n =

0 for all other times. Suppose further that we sample the signal with a sampling frequency f ;. We can write x # n = PSD, Autocorrelation, and Noise in MATLAB Aaron Scher Energy and power of a signal Consider a continuous time deterministic signal x t. We are specifically interested in analyzing the characteristics of

More information

Statistics Homework #4

Statistics Homework #4 Statistics 910 1 Homework #4 Chapter 6, Shumway and Stoffer These are outlines of the solutions. If you would like to fill in other details, please come see me during office hours. 6.1 State-space representation

More information

Discrete-Time Gaussian Fourier Transform Pair, and Generating a Random Process with Gaussian PDF and Power Spectrum Mark A.

Discrete-Time Gaussian Fourier Transform Pair, and Generating a Random Process with Gaussian PDF and Power Spectrum Mark A. Discrete-Time Gaussian Fourier Transform Pair, and Generating a Random Process with Gaussian PDF and Power Spectrum Mark A. Richards October 3, 6 Updated April 5, Gaussian Transform Pair in Continuous

More information

Laboratory handout 5 Mode shapes and resonance

Laboratory handout 5 Mode shapes and resonance laboratory handouts, me 34 82 Laboratory handout 5 Mode shapes and resonance In this handout, material and assignments marked as optional can be skipped when preparing for the lab, but may provide a useful

More information

7.17. Determine the z-transform and ROC for the following time signals: Sketch the ROC, poles, and zeros in the z-plane. X(z) = x[n]z n.

7.17. Determine the z-transform and ROC for the following time signals: Sketch the ROC, poles, and zeros in the z-plane. X(z) = x[n]z n. Solutions to Additional Problems 7.7. Determine the -transform and ROC for the following time signals: Sketch the ROC, poles, and eros in the -plane. (a) x[n] δ[n k], k > 0 X() x[n] n n k, 0 Im k multiple

More information

Computer Exercise 0 Simulation of ARMA-processes

Computer Exercise 0 Simulation of ARMA-processes Lund University Time Series Analysis Mathematical Statistics Fall 2018 Centre for Mathematical Sciences Computer Exercise 0 Simulation of ARMA-processes The purpose of this computer exercise is to illustrate

More information